09. Mini Project: Training an MLP on MNIST
Mini Project: Training an MLP on MNIST
Deep learning is not well-understood, and the practice is ahead of the theory in many cases. If you are new to deep learning, you are strongly encouraged to experiment with many models, to develop intuition about why models work. Starter code is provided on github.
In this mini project, you'll modify the neural network in mnist_mlp.ipynb and compare the resultant model configurations.
Remember: Overfitting is detected by comparing the validation loss to the training loss. If the training loss is much lower than the validation loss, then the model might be overfitting.
Instructions
Task Description:
Take note of the validation loss and test accuracy for the model resulting from each of the below changes. (Try out each amendment in a separate model.)
Task Feedback:
Great job! Now, you're ready to learn about Convolutional Neural Networks (CNNs)!
Optional Resource
If you're interested in learning how to do a more systematic hyperparameter search in your neural networks, please check out this blog post.